lexical-core
Low-level, lexical conversion routines for use in a no_std
context. This crate by default does not use the Rust standard library.
- Getting Started
- Features
- Configuration
- Constants
- Documentation
- Validation
- Implementation Details
- Known Issues
- Version Support
- Changelog
- License
- Contributing
Getting Started
lexical-core is a low-level API for number-to-string and string-to-number conversions, without requiring a system allocator. If you would like to use a convenient, high-level API, please look at lexical instead.
Add lexical-core to your Cargo.toml
:
[]
= "^0.6.7"
And an introduction through use:
extern crate lexical_core;
// String to number using Rust slices.
// The argument is the byte string parsed.
let f: f32 = parse.unwrap; // 3.5
let i: i32 = parse.unwrap; // 15
// All lexical_core parsers are checked, they validate the
// input data is entirely correct, and stop parsing when invalid data
// is found, or upon numerical overflow.
let r = ; // Err(ErrorCode::Overflow.into())
let r = ; // Err(ErrorCode::InvalidDigit.into())
// In order to extract and parse a number from a substring of the input
// data, use `parse_partial`. These functions return the parsed value and
// the number of processed digits, allowing you to extract and parse the
// number in a single pass.
let r = ; // Ok((3, 1))
// If an insufficiently long buffer is passed, the serializer will panic.
// PANICS
let mut buf = ;
//let slc = lexical_core::write::<i64>(15, &mut buf);
// In order to guarantee the buffer is long enough, always ensure there
// are at least `T::FORMATTED_SIZE` bytes, which requires the
// `lexical_core::Number` trait to be in scope.
use Number;
let mut buf = ;
let slc = ;
assert_eq!;
// When the `radix` feature is enabled, for decimal floats, using
// `T::FORMATTED_SIZE` may significantly overestimate the space
// required to format the number. Therefore, the
// `T::FORMATTED_SIZE_DECIMAL` constants allow you to get a much
// tighter bound on the space required.
let mut buf = ;
let slc = ;
assert_eq!;
Features
- correct Use a correct string-to-float parser.
- trim_floats Export floats without a fraction as an integer.
- radix Allow conversions to and from non-decimal strings.
- format Customize accepted inputs for number parsing.
- rounding Enable custom rounding for IEEE754 floats.
- ryu Use dtolnay's ryu library for float-to-string conversions.
Format
Every language has competing specifications for valid numerical input, meaning a number parser for Rust will incorrectly accept or reject input for different programming or data languages. For example:
extern crate lexical_core;
use *;
// Valid in Rust strings.
// Not valid in JSON.
let f: f64 = parse.unwrap; // 3e7
// Let's only accept JSON floats.
let format = JSON;
let f: f64 = parse_format.unwrap; // 3e7
let f: f64 = parse_format.unwrap; // Panics!
// We can also allow digit separators, for example.
// OCaml, a programming language that inspired Rust,
// accepts digit separators pretty much anywhere.
let format = OCAML_STRING;
let f: f64 = parse.unwrap; // Panics!
let f: f64 = parse_format.unwrap; // 34.01
The parsing specification is defined by NumberFormat
, which provides pre-defined constants for over 40 programming and data languages. However, it also allows you to create your own specification, to dictate parsing.
extern crate lexical_core;
use *;
// Let's use the standard, Rust grammar.
let format = standard.unwrap;
// Let's use a permissive grammar, one that allows anything besides
// digit separators.
let format = permissive.unwrap;
// Let's ignore digit separators and have an otherwise permissive grammar.
let format = ignore.unwrap;
// Create our own grammar.
// A NumberFormat is compiled from options into binary flags, each
// taking 1-bit, allowing high-performance, customizable parsing
// once they're compiled. Each flag will be explained while defining it.
// The '_' character will be used as a digit separator.
let digit_separator = b'_';
// Require digits in the integer component of a float.
// `0.1` is valid, but `.1` is not.
let required_integer_digits = false;
// Require digits in the fraction component of a float.
// `1.0` is valid, but `1.` and `1` are not.
let required_fraction_digits = false;
// Require digits in the exponent component of a float.
// `1.0` and `1.0e7` is valid, but `1.0e` is not.
let required_exponent_digits = false;
// Do not allow a positive sign before the mantissa.
// `1.0` and `-1.0` are valid, but `+1.0` is not.
let no_positive_mantissa_sign = false;
// Require a sign before the mantissa.
// `+1.0` and `-1.0` are valid, but `1.0` is not.
let required_mantissa_sign = false;
// Do not allow the use of exponents.
// `300.0` is valid, but `3.0e2` is not.
let no_exponent_notation = false;
// Do not allow a positive sign before the exponent.
// `3.0e2` and 3.0e-2` are valid, but `3.0e+2` is not.
let no_positive_exponent_sign = false;
// Require a sign before the exponent.
// `3.0e+2` and `3.0e-2` are valid, but `3.0e2` is not.
let required_exponent_sign = false;
// Do not allow an exponent without fraction digits.
// `3.0e7` is valid, but `3e7` and `3.e7` are not.
let no_exponent_without_fraction = false;
// Do not allow special values.
// `1.0` is valid, but `NaN` and `inf` are not.
let no_special = false;
// Use case-sensitive matching when parsing special values.
// `NaN` is valid, but `nan` and `NAN` are not.
let case_sensitive_special = false;
// Allow digit separators between digits in the integer component.
// `3_4.01` is valid, but `_34.01`, `34_.01` and `34.0_1` are not.
let integer_internal_digit_separator = false;
// Allow digit separators between digits in the fraction component.
// `34.0_1` is valid, but `34._01`, `34.01_` and `3_4.01` are not.
let fraction_internal_digit_separator = false;
// Allow digit separators between digits in the exponent component.
// `1.0e6_7` is valid, but `1.0e_67`, `1.0e67_` and `1_2.0e67` are not.
let exponent_internal_digit_separator = false;
// Allow digit separators before any digits in the integer component.
// These digit separators may occur before or after the sign, as long
// as they occur before any digits.
// `_34.01` is valid, but `3_4.01`, `34_.01` and `34._01` are not.
let integer_leading_digit_separator = false;
// Allow digit separators before any digits in the fraction component.
// `34._01` is valid, but `34.0_1`, `34.01_` and `_34.01` are not.
let fraction_leading_digit_separator = false;
// Allow digit separators before any digits in the exponent component.
// These digit separators may occur before or after the sign, as long
// as they occur before any digits.
// `1.0e_67` is valid, but `1.0e6_7`, `1.0e67_` and `_1.0e67` are not.
let exponent_leading_digit_separator = false;
// Allow digit separators after any digits in the integer component.
// If `required_integer_digits` is not set, `_.01` is valid.
// `34_.01` is valid, but `3_4.01`, `_34.01` and `34.01_` are not.
let integer_trailing_digit_separator = false;
// Allow digit separators after any digits in the fraction component.
// If `required_fraction_digits` is not set, `1._` is valid.
// `34.01_` is valid, but `34.0_1`, `34._01` and `34_.01` are not.
let fraction_trailing_digit_separator = false;
// Allow digit separators after any digits in the exponent component.
// If `required_exponent_digits` is not set, `1.0e_` is valid.
// `1.0e67_` is valid, but `1.0e6_7`, `1.0e_67` and `1.0_e67` are not.
let exponent_trailing_digit_separator = false;
// Allow consecutive separators in the integer component.
// This requires another integer digit separator flag to be set.
// For example, if `integer_internal_digit_separator` and this flag are set,
// `3__4.01` is valid, but `__34.01`, `34__.01` and `34.0__1` are not.
let integer_consecutive_digit_separator = false;
// Allow consecutive separators in the fraction component.
// This requires another fraction digit separator flag to be set.
// For example, if `fraction_internal_digit_separator` and this flag are set,
// `34.0__1` is valid, but `34.__01`, `34.01__` and `3__4.01` are not.
let fraction_consecutive_digit_separator = false;
// Allow consecutive separators in the exponent component.
// This requires another exponent digit separator flag to be set.
// For example, if `exponent_internal_digit_separator` and this flag are set,
// `1.0e6__7` is valid, but `1.0e__67`, `1.0e67__` and `1__2.0e67` are not.
let exponent_consecutive_digit_separator = false;
// Allow digit separators in special values.
// If set, allow digit separators in special values will be ignored.
// `N_a_N__` is valid, but `i_n_f_e` is not.
let special_digit_separator = false;
// Compile the grammar.
let format = compile.unwrap;
Configuration
Lexical-core also includes configuration options that allow you to configure float processing and formatting. These are provided as getters and setters, so lexical-core can validate the input.
- NaN
get_nan_string
set_nan_string
- Short Infinity
get_inf_string
set_inf_string
- Long Infinity
get_infinity_string
set_infinity_string
- Exponent Default Character
get_exponent_default_char
set_exponent_default_char
- Exponent Backup Character (radix only)
get_exponent_backup_char
set_exponent_backup_char
- Float Rounding (rounding only)
get_float_rounding
set_float_rounding
Constants
Lexical-core also includes a few constants to simplify interfacing with number-to-string code, and are implemented for the lexical_core::Number
trait, which is required by ToLexical
.
- FORMATTED_SIZE The maximum number of bytes a formatter may write.
- FORMATTED_SIZE_DECIMAL The maximum number of bytes a formatter may write in decimal (base 10).
These are provided as Rust constants so they may be used as the size element in arrays.
Documentation
Lexical-core's documentation can be found on docs.rs.
Validation
Float parsing is difficult to do correctly, and major bugs have been found in implementations from libstdc++'s strtod to Python. In order to validate the accuracy of the lexical, we employ the following external tests:
- Hrvoje Abraham's strtod test cases.
- Rust's test-float-parse unittests.
- Testbase's stress tests for converting from decimal to binary.
- Various difficult cases reported on blogs.
Although lexical may contain bugs leading to rounding error, it is tested against a comprehensive suite of random-data and near-halfway representations, and should be fast and correct for the vast majority of use-cases.
Implementation Details
Float to String
For more information on the Grisu2 and Grisu3 algorithms, see Printing Floating-Point Numbers Quickly and Accurately with Integers.
For more information on the Ryu algorithm, see Ryƫ: fast float-to-string conversion.
String to Float
In order to implement an efficient parser in Rust, lexical uses the following steps:
- We ignore the sign until the end, and merely toggle the sign bit after creating a correct representation of the positive float.
- We handle special floats, such as "NaN", "inf", "Infinity". If we do not have a special float, we continue to the next step.
- We parse up to 64-bits from the string for the mantissa, ignoring any trailing digits, and parse the exponent (if present) as a signed 32-bit integer. If the exponent overflows or underflows, we set the value to i32::max_value() or i32::min_value(), respectively.
- Fast Path We then try to create an exact representation of a native binary float from parsed mantissa and exponent. If both can be exactly represented, we multiply the two to create an exact representation, since IEEE754 floats mandate the use of guard digits to minimizing rounding error. If either component cannot be exactly represented as the native float, we continue to the next step.
- Moderate Path We create an approximate, extended, 80-bit float type (64-bits for the mantissa, 16-bits for the exponent) from both components, and multiplies them together. This minimizes the rounding error, through guard digits. We then estimate the error from the parsing and multiplication steps, and if the float +/- the error differs significantly from b+h, we return the correct representation (b or b+u). If we cannot unambiguously determine the correct floating-point representation, we continue to the next step.
- Fallback Moderate Path Next, we create a 128-bit representation of the numerator and denominator for b+h, to disambiguate b from b+u by comparing the actual digits in the input to theoretical digits generated from b+h. This is accurate for ~36 significant digits from a 128-bit approximation with decimal float strings. If the input is less than or equal to 36 digits, we return the value from this step. Otherwise, we continue to the next step.
- Slow Path We use arbitrary-precision arithmetic to disambiguate the correct representation without any rounding error. We create an exact representation of the input digits as a big integer, to determine how to round the top 53 bits for the mantissa. If there is a fraction or a negative exponent, we create a representation of the significant digits for
b+h
and scale the input digits by the binary exponent inb+h
, and scale the significant digits inb+h
by the decimal exponent, and compare the two to determine if we need to round up or down.
Since arbitrary-precision arithmetic is slow and scales poorly for decimal strings with many digits or exponents of high magnitude, lexical also supports a lossy algorithm, which returns the result from the moderate path. The result from the lossy parser should be accurate to within 1 ULP.
Arbitrary-Precision Arithmetic
Lexical uses arbitrary-precision arithmetic to exactly represent strings between two floating-point representations, and is highly optimized for performance. The following section is a comparison of different algorithms to determine the correct float representation. The arbitrary-precision arithmetic logic is not dependent on memory allocation: it only uses the heap when the radix
feature is enabled.
Algorithm Background and Comparison
For close-to-halfway representations of a decimal string s
, where s
is close between two representations, b
and the next float b+u
, arbitrary-precision arithmetic is used to determine the correct representation. This means s
is close to b+h
, where h
is the halfway point between b
and b+u
.
For the following example, we will use the following values for our test case:
s = 2.4703282292062327208828439643411068618252990130716238221279284125033775363510437593264991818081799618989828234772285886546332835517796989819938739800539093906315035659515570226392290858392449105184435931802849936536152500319370457678249219365623669863658480757001585769269903706311928279558551332927834338409351978015531246597263579574622766465272827220056374006485499977096599470454020828166226237857393450736339007967761930577506740176324673600968951340535537458516661134223766678604162159680461914467291840300530057530849048765391711386591646239524912623653881879636239373280423891018672348497668235089863388587925628302755995657524455507255189313690836254779186948667994968324049705821028513185451396213837722826145437693412532098591327667236328125001e-324
b = 0.0
b+h = 2.4703282292062327208828439643411068618252990130716238221279284125033775363510437593264991818081799618989828234772285886546332835517796989819938739800539093906315035659515570226392290858392449105184435931802849936536152500319370457678249219365623669863658480757001585769269903706311928279558551332927834338409351978015531246597263579574622766465272827220056374006485499977096599470454020828166226237857393450736339007967761930577506740176324673600968951340535537458516661134223766678604162159680461914467291840300530057530849048765391711386591646239524912623653881879636239373280423891018672348497668235089863388587925628302755995657524455507255189313690836254779186948667994968324049705821028513185451396213837722826145437693412532098591327667236328125e-324
b+u = 5e-324
Algorithm M
Algorithm M represents the significant digits of a float as a fraction of arbitrary-precision integers (a more in-depth description can be found here). For example, 1.23 would be 123/100, while 314.159 would be 314159/1000. We then scale the numerator and denominator by powers of 2 until the quotient is in the range [2^52, 2^53)
, generating the correct significant digits of the mantissa.
A naive implementation, in Python, is as follows:
# Ensure numerator >= 2**52
=
<<= 53
-= 53
# Track number of steps required (optional).
= 0
+= 1
= //
//= 2
*= 2
break
return
bigcomp
Bigcomp is the canonical string-to-float parser, which creates an exact representation of b+h
as a big integer, and compares the theoretical digits from b+h
scaled into the range [1, 10)
by a power of 10 to the actual digits in the input string (a more in-depth description can be found here). A maximum of 768 digits need to be compared to determine the correct representation, and the size of the big integers in the ratio does not depend on the number of digits in the input string.
Bigcomp is used as a fallback algorithm for lexical-core when the radix feature is enabled, since the radix-representation of a binary float may never terminate if the radix is not divisible by 2. Since bigcomp uses constant memory, it is used as the default algorithm if more than 2^15
digits are passed and the representation is potentially non-terminating.
bhcomp
Bhcomp is a simple, performant algorithm that compared the significant digits to the theoretical significant digits for b+h
. Simply, the significant digits from the string are parsed, creating a ratio. A ratio is generated for b+h
, and these two ratios are scaled using the binary and radix exponents.
For example, "2.470328e-324" produces a ratio of 2470328/10^329
, while b+h
produces a binary ratio of 1/2^1075
. We're looking to compare these ratios, so we need to scale them using common factors. Here, we convert this to (2470328*5^329*2^1075)/10^329
and (1*5^329*2^1075)/2^1075
, which converts to 2470328*2^746
and 1*5^329
.
Our significant digits (real_digits) and b+h
(bh_digits) therefore start like:
real_digits = 91438982...
bh_digits = 91438991...
Since our real digits are below the theoretical halfway point, we know we need to round-down, meaning our literal value is b
, or 0.0
. This approach allows us to calculate whether we need to round-up or down with a single comparison step, without any native divisions required. This is the default algorithm lexical-core uses.
Other Optimizations
- We remove powers of 2 during exponentiation in bhcomp.
- We limit the number of parsed digits to the theoretical max number of digits produced by
b+h
(768 for decimal strings), and merely compare any trailing digits to '0'. This provides an upper-bound on the computation cost. - We use fast exponentiation and multiplication algorithms to scale the significant digits for comparison.
- For the fallback bigcomp algorithm, we use a division algorithm optimized for the generation of a single digit from a given radix, by setting the leading bit in the denominator 4 below the most-significant bit (in decimal strings). This requires only 1 native division per digit generated.
- The individual "limbs" of the big integers are optimized to the architecture we compile on, for example, u32 on x86 and u64 on x86-64, minimizing the number of native operations required. Currently, 64-bit limbs are used on target architectures
aarch64
,powerpc64
,mips64
, andx86_64
.
Known Issues
On the ARMVv6 architecture, the stable exponentiation for the fast, incorrect float parser is not fully stable. For example, 1e-300
is correct, while 5e-324
rounds to 0
, leading to "5e-324" being incorrectly parsed as 0
. This does not affect the default, correct float parser, nor ARMVv7 or ARMVv8 (aarch64) architectures. This bug can compound errors in the incorrect parser (feature-gated by disabling the correct
feature`). It is not known if this bug is an artifact of Qemu emulation of ARMv6, or is actually representative the hardware.
Versions of lexical-core prior to 0.4.3 could round parsed floating-point numbers with an error of up to 1 ULP. This occurred for strings with 16 or more digits and a trailing 0 in the fraction, the b+h
comparison in the slow-path algorithm incorrectly scales the the theoretical digits due to an over-calculated real exponent. This affects a very small percentage of inputs, however, it is recommended to update immediately.
Version Support
Lexical-core is tested to work from Rustc versions of 1.31.0-1.51.0, and should work on newer versions as well. Please report any errors compiling lexical-core for any Rust compiler 1.31.0 or later. Please note the test suite require a Rustc version of 1.36 or later.
Changelog
All changes since 2.2.0 are documented in CHANGELOG.
License
Lexical-core is dual licensed under the Apache 2.0 license as well as the MIT license. See the LICENCE-MIT and the LICENCE-APACHE files for the licenses.
Lexical-core also ports some code from rust (for backwards compatibility), V8, libgo and fpconv, and therefore might be subject to the terms of a 3-clause BSD license or BSD-like license.
Contributing
Unless you explicitly state otherwise, any contribution intentionally submitted for inclusion in lexical by you, as defined in the Apache-2.0 license, shall be dual licensed as above, without any additional terms or conditions.